randomized trial
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > United States > Minnesota (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
Machine learning to optimize precision in the analysis of randomized trials: A journey in pre-specified, yet data-adaptive learning
Balzer, Laura B., van der Laan, Mark J., Petersen, Maya L.
Covariate adjustment is an approach to improve the precision of trial analyses by adjusting for baseline variables that are prognostic of the primary endpoint. Motivated by the SEARCH Universal HIV Test-and-Treat Trial (2013-2017), we tell our story of developing, evaluating, and implementing a machine learning-based approach for covariate adjustment. We provide the rationale for as well as the practical concerns with such an approach for estimating marginal effects. Using schematics, we illustrate our procedure: targeted machine learning estimation (TMLE) with Adaptive Pre-specification. Briefly, sample-splitting is used to data-adaptively select the combination of estimators of the outcome regression (i.e., the conditional expectation of the outcome given the trial arm and covariates) and known propensity score (i.e., the conditional probability of being randomized to the intervention given the covariates) that minimizes the cross-validated variance estimate and, thereby, maximizes empirical efficiency. We discuss our approach for evaluating finite sample performance with parametric and plasmode simulations, pre-specifying the Statistical Analysis Plan, and unblinding in real-time on video conference with our colleagues from around the world. We present the results from applying our approach in the primary, pre-specified analysis of 8 recently published trials (2022-2024). We conclude with practical recommendations and an invitation to implement our approach in the primary analysis of your next trial.
- North America > United States > California > Alameda County > Berkeley (0.14)
- Africa > Uganda (0.06)
- Africa > Kenya (0.06)
- (8 more...)
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (1.00)
- Health & Medicine > Therapeutic Area > Immunology > HIV (0.72)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.49)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.49)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.47)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > United States > Minnesota (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
Robust estimation of heterogeneous treatment effects in randomized trials leveraging external data
Karlsson, Rickard, De Bartolomeis, Piersilvio, Dahabreh, Issa J., Krijthe, Jesse H.
Randomized trials are typically designed to detect average treatment effects but often lack the statistical power to uncover effect heterogeneity over patient characteristics, limiting their value for personalized decision-making. To address this, we propose the QR-learner, a model-agnostic learner that estimates conditional average treatment effects (CATE) within the trial population by leveraging external data from other trials or observational studies. The proposed method is robust: it has the potential to reduce the CATE prediction mean squared error while maintaining consistency, even when the external data is not aligned with the trial. Moreover, we introduce a procedure that combines the QR-learner with a trial-only CATE learner and show that it asymptotically matches or exceeds the trial-only learner in terms of mean squared error. We examine the performance of our approach in simulation studies and apply the methods to a real-world dataset, demonstrating improvements in both CATE estimation and statistical power for detecting heterogeneous effects.
- Europe > Netherlands > South Holland > Delft (0.04)
- North America > United States > Tennessee (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Research Report > Strength High (1.00)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Education (1.00)
- Government > Regional Government > North America Government > United States Government (0.46)
- Health & Medicine > Public Health (0.46)
Admissibility of Completely Randomized Trials: A Large-Deviation Approach
Imbens, Guido, Qin, Chao, Wager, Stefan
When an experimenter has the option of running an adaptive trial, is it admissible to ignore this option and run a non-adaptive trial instead? We provide a negative answer to this question in the best-arm identification problem, where the experimenter aims to allocate measurement efforts judiciously to confidently deploy the most effective treatment arm. We find that, whenever there are at least three treatment arms, there exist simple adaptive designs that universally and strictly dominate non-adaptive completely randomized trials. This dominance is characterized by a notion called efficiency exponent, which quantifies a design's statistical efficiency when the experimental sample is large. Our analysis focuses on the class of batched arm elimination designs, which progressively eliminate underperforming arms at pre-specified batch intervals. We characterize simple sufficient conditions under which these designs universally and strictly dominate completely randomized trials. These results resolve the second open problem posed in Qin [2022].
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > Jordan (0.04)
Predicting and preventing Alzheimer's disease Science
With all the advances in both the science of aging and artificial intelligence (AI), we are in a propitious position to accurately and precisely determine who is at high risk of developing Alzheimer's disease years before signs of even mild cognitive deficit. It takes at least 20 years for aggregates of misfolded β-amyloid and tau proteins to accumulate in the brain along with neuroinflammation that they incite. This provides a long window of opportunity to get ahead of the pathobiological process, both for prediction and prevention. A family history of Alzheimer's and the presence of genetic variants in the APOE4 (apolipoprotein E4) allele indicate an increased risk, as does a polygenic risk score that is based on the combined influence of many genetic variants. However, each of these clues provides little insight about when initial symptoms would likely present.
Detecting critical treatment effect bias in small subgroups
De Bartolomeis, Piersilvio, Abad, Javier, Donhauser, Konstantin, Yang, Fanny
Randomized trials have traditionally been the gold standard for informed decision-making in medicine, as they allow for unbiased estimation of treatment effects under mild assumptions. However, there is often a significant discrepancy between the patients observed in clinical practice and those enrolled in randomized trials, limiting the generalizability of the trial results [12, 43]. To address this issue, the U.S. Food and Drug Administration advocates for using observational data, as it is usually more representative of the patient population in clinical practice [30, 39]. Yet, a major caveat to this recommendation is that several sources of bias, including hidden confounding, can compromise the causal conclusions drawn from observational data. In light of the inherent limitations of randomized and observational data, it has become a popular strategy to benchmark observational studies against existing randomized trials to assess their quality [4, 13]. The main idea behind this approach is first to emulate the procedures adopted in the randomized trial within the observational study; see e.g.
- North America > United States > Virginia (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)
- Information Technology > Data Science (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.68)
Hidden yet quantifiable: A lower bound for confounding strength using randomized trials
De Bartolomeis, Piersilvio, Abad, Javier, Donhauser, Konstantin, Yang, Fanny
In the era of fast-paced precision medicine, observational studies play a major role in properly evaluating new treatments in clinical practice. Yet, unobserved confounding can significantly compromise causal conclusions drawn from non-randomized data. We propose a novel strategy that leverages randomized trials to quantify unobserved confounding. First, we design a statistical test to detect unobserved confounding with strength above a given threshold. Then, we use the test to estimate an asymptotically valid lower bound on the unobserved confounding strength. We evaluate the power and validity of our statistical test on several synthetic and semi-synthetic datasets. Further, we show how our lower bound can correctly identify the absence and presence of unobserved confounding in a real-world setting.
- North America > United States > Tennessee (0.04)
- North America > United States > Michigan (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)
Adaptive Selection of the Optimal Strategy to Improve Precision and Power in Randomized Trials
Balzer, Laura B., Cai, Erica, Garraza, Lucas Godoy, Amaranath, Pracheta
Benkeser et al. demonstrate how adjustment for baseline covariates in randomized trials can meaningfully improve precision for a variety of outcome types. Their findings build on a long history, starting in 1932 with R.A. Fisher and including more recent endorsements by the U.S. Food and Drug Administration and the European Medicines Agency. Here, we address an important practical consideration: *how* to select the adjustment approach -- which variables and in which form -- to maximize precision, while maintaining Type-I error control. Balzer et al. previously proposed *Adaptive Prespecification* within TMLE to flexibly and automatically select, from a prespecified set, the approach that maximizes empirical efficiency in small trials (N$<$40). To avoid overfitting with few randomized units, selection was previously limited to working generalized linear models, adjusting for a single covariate. Now, we tailor Adaptive Prespecification to trials with many randomized units. Using $V$-fold cross-validation and the estimated influence curve-squared as the loss function, we select from an expanded set of candidates, including modern machine learning methods adjusting for multiple covariates. As assessed in simulations exploring a variety of data generating processes, our approach maintains Type-I error control (under the null) and offers substantial gains in precision -- equivalent to 20-43\% reductions in sample size for the same statistical power. When applied to real data from ACTG Study 175, we also see meaningful efficiency improvements overall and within subgroups.
- North America > United States > New York (0.04)
- Europe > Netherlands > South Holland > Dordrecht (0.04)
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)
Improved Policy Evaluation for Randomized Trials of Algorithmic Resource Allocation
Mate, Aditya, Wilder, Bryan, Taneja, Aparna, Tambe, Milind
We consider the task of evaluating policies of algorithmic resource allocation through randomized controlled trials (RCTs). Such policies are tasked with optimizing the utilization of limited intervention resources, with the goal of maximizing the benefits derived. Evaluation of such allocation policies through RCTs proves difficult, notwithstanding the scale of the trial, because the individuals' outcomes are inextricably interlinked through resource constraints controlling the policy decisions. Our key contribution is to present a new estimator leveraging our proposed novel concept, that involves retrospective reshuffling of participants across experimental arms at the end of an RCT. We identify conditions under which such reassignments are permissible and can be leveraged to construct counterfactual trials, whose outcomes can be accurately ascertained, for free. We prove theoretically that such an estimator is more accurate than common estimators based on sample means -- we show that it returns an unbiased estimate and simultaneously reduces variance. We demonstrate the value of our approach through empirical experiments on synthetic, semi-synthetic as well as real case study data and show improved estimation accuracy across the board.
- North America > United States (0.04)
- Asia > India > Maharashtra > Mumbai (0.04)
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)